# Chat optimization
GLM 4 32B 0414 4bit DWQ
MIT
This is the MLX format version of the THUDM/GLM-4-32B-0414 model, processed with 4-bit DWQ quantization, suitable for efficient inference on Apple silicon devices.
Large Language Model Supports Multiple Languages
G
mlx-community
156
4
Josiefied Qwen3 8B Abliterated V1 8bit
An optimized 8-bit quantized version of Qwen3-8B, designed for efficient inference on the MLX framework
Large Language Model
J
mlx-community
450
1
Qwen2.5 7B Instruct Quantized.w8a8
Apache-2.0
INT8 quantized version of Qwen2.5-7B-Instruct, suitable for multilingual scenarios in both commercial and research applications, optimized for memory requirements and computational throughput.
Large Language Model
Safetensors English
Q
RedHatAI
412
1
H2o Danube3 4b Chat
Apache-2.0
A chat model fine-tuned by H2O.ai with 4 billion parameters, supporting text generation tasks and capable of offline operation on mobile phones.
Large Language Model
Transformers English

H
h2oai
41.49k
67
Leniachat Gemma 2b V0
Apache-2.0
A 2B-parameter chat and instruction model optimized for Spanish, supporting text generation, dialogue, and Q&A tasks
Large Language Model
Transformers Spanish

L
LenguajeNaturalAI
132
14
Featured Recommended AI Models